Discovery of Latent Factors in High-dimensional Data Using Tensor Methods
نویسنده
چکیده
OF THE DISSERTATIONDiscovery of Latent Factors in High-dimensional Data Using Tensor MethodsByFurong HuangDoctor of Philosophy in Electrical and Computer EngineeringUniversity of California, Irvine, 2016Assistant Professor Animashree Anandkumar, Chair Unsupervised learning aims at the discovery of hidden structure that drives the observationsin the real world. It is essential for success in modern machine learning and artificial intel-ligence. Latent variable models are versatile in unsupervised learning and have applicationsin almost every domain, e.g., social network analysis, natural language processing, computervision and computational biology. Training latent variable models is challenging due to thenon-convexity of the likelihood objective function. An alternative method is based on thespectral decomposition of low order moment matrices and tensors. This versatile frameworkis guaranteed to estimate the correct model consistently. My thesis spans both theoreticalanalysis of tensor decomposition framework and practical implementation of various appli-cations. This thesis presents theoretical results on convergence to globally optimal solution of tensordecomposition using the stochastic gradient descent, despite non-convexity of the objective.This is the first work that gives global convergence guarantees for the stochastic gradientdescent on non-convex functions with exponentially many local minima and saddle points. This thesis also presents large-scale deployment of spectral methods (matrix and tensordecomposition) carried out on CPU, GPU and Spark platforms. Dimensionality reductiontechniques such as random projection are incorporated for a highly parallel and scalablexvi tensor decomposition algorithm. We obtain a gain in both accuracies and in running timesby several orders of magnitude compared to the state-of-art variational methods. To solve real world problems, more advanced models and learning algorithms are proposed.After introducing tensor decomposition framework under latent Dirichlet allocation (LDA)model, this thesis discusses generalization of LDA model to mixed membership stochasticblock model for learning hidden user commonalities or communities in social network, con-volutional dictionary model for learning phrase templates and word-sequence embeddings,hierarchical tensor decomposition and latent tree structure model for learning disease hierar-chy in healthcare analytics, and spatial point process mixture model for detecting cell typesin neuroscience.
منابع مشابه
Distributed Latent Dirichlet Allocation on Spark via Tensor Decomposition
Learning latent variable mixture models in high-dimension is applicable in numerous domains where low dimensional latent factors out of the high-dimensional observations are desired. Popular likelihood based methods optimize over a non-convex likelihood which is computationally challenging to achieve due to the high-dimensionality of the data, and therefore it is usually not guaranteed to conve...
متن کاملNumerical methods in higher dimensions using tensor factorizations
Numerical methods in higher dimensions using tensor factorizations I this talk I will collect recent advances in the solution of high-dimensional problems in different application areas: chemistry, biology, mathematics. The language of low-rank factorization gives a unified view on different algorithms for the solution of seemingly diverse and unconnected problems. Typical applications include ...
متن کاملNeural Tensor Factorization
Neural collaborative filtering (NCF) [15] and recurrent recommender systems (RRN) [38] have been successful in modeling user-item relational data. However, they are also limited in their assumption of static or sequential modeling of relational data as they do not account for evolving users’ preference over time as well as changes in the underlying factors that drive the change in user-item rel...
متن کاملGaussian process based nonlinear latent structure discovery in multivariate spike train data
A large body of recent work focuses on methods for extracting low-dimensional latent structure from multi-neuron spike train data. Most such methods employ either linear latent dynamics or linear mappings from latent space to log spike rates. Here we propose a doubly nonlinear latent variable model that can identify low-dimensional structure underlying apparently high-dimensional spike train da...
متن کاملGaussian process based nonlinear latent structure discovery in multivariate spike train data
A large body of recent work has focused on methods for identifying lowdimensional latent structure in multi-neuron spike train data. Most such methods have employed either linear latent dynamics or linear (or log-linear) mappings from a latent space to spike rates. Here we propose a doubly nonlinear latent variable model for population spike trains that can identify nonlinear low-dimensional st...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1606.03212 شماره
صفحات -
تاریخ انتشار 2016